Nearest Neighbor Classi cation from Multiple Feature Subsets

نویسنده

  • Stephen D. Bay
چکیده

Combining multiple classi ers is an e ective technique for improving accuracy. There are many general combining algorithms, such as Bagging, Boosting, or Error Correcting Output Coding, that signi cantly improve classi ers like decision trees, rule learners, or neural networks. Unfortunately, these combining methods do not improve the nearest neighbor classi er. In this paper, we present MFS, a combining algorithm designed to improve the accuracy of the nearest neighbor (NN) classi er. MFS combines multiple NN classi ers each using only a random subset of features. The experimental results are encouraging: On 25 datasets from the UCI Repository, MFS signi cantly outperformed several standard NN variants and was competitive with boosted decision trees. In additional experiments, we show that MFS is robust to irrelevant features, and is able to reduce both bias and variance components of error.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Combining Nearest Neighbor Classifiers Through Multiple Feature Subsets

Combining multiple classi ers is an e ective technique for improving accuracy. There are many general combining algorithms, such as Bagging or Error Correcting Output Coding, that signi cantly improve classi ers like decision trees, rule learners, or neural networks. Unfortunately, many combining methods do not improve the nearest neighbor classi er. In this paper, we present MFS, a combining a...

متن کامل

Application of k Nearest Neighbor on Feature Projections Classi er to Text Categorization

This paper presents the results of the application of an instance based learning algorithm k Nearest Neighbor Method on Fea ture Projections k NNFP to text categorization and compares it with k Nearest Neighbor Classi er k NN k NNFP is similar to k NN ex cept it nds the nearest neighbors according to each feature separately Then it combines these predictions using a majority voting This prop er...

متن کامل

Cloud Classi cation Using Error-Correcting Output Codes

Novel arti cial intelligence methods are used to classify 16x16 pixel regions (obtained from Advanced Very High Resolution Radiometer (AVHRR) images) in terms of cloud type (e.g., stratus, cumulus, etc.). We previously reported that intelligent feature selection methods, combined with nearest neighbor classi ers, can dramatically improve classi cation accuracy on this task. Our subsequent analy...

متن کامل

DeEPs: A New Instance-based Discovery and Classi cation System

Distance is used as a measurement in most of instance-based classi cation systems. Rather than using distance, we make use of the frequency of an instance's subsets and the frequency-changing rate of the subsets among training classes to perform both knowledge discovery and classi cation tasks. We name the system DeEPs. Whenever an instance is considered, DeEPs can e ciently discover those patt...

متن کامل

Symbolic Nearest Mean Classiiers

Piew Datta and Dennis Kibler Department of Information and Computer Science University of California Irvine, CA 92717 fpdatta, [email protected] Abstract The minimum-distance classi er summarizes each class with a prototype and then uses a nearest neighbor approach for classi cation. Three drawbacks of the minimum-distance classi er are its inability to work with symbolic attributes, weigh at...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1999